addHessian: Combining quasi-Newton method with first-order method for neural network training
نویسندگان
چکیده
First-order methods such as SGD and Adam are popularly used in training Neural networks. On the other hand, second-order have shown to better performance faster convergence despite their high computational cost by incorporating curvature information. While determine step size line search approaches, first-order achieve efficient learning devising a way adjust size. In this paper, we propose new algorithm for neural networks combining methods. We investigate effectiveness of our proposed method when combined with popular - SGD, Adagrad, Adam, through experiments using image classification problems.
منابع مشابه
A conjugate gradient based method for Decision Neural Network training
Decision Neural Network is a new approach for solving multi-objective decision-making problems based on artificial neural networks. Using inaccurate evaluation data, network training has improved and the number of educational data sets has decreased. The available training method is based on the gradient decent method (BP). One of its limitations is related to its convergence speed. Therefore,...
متن کاملTraining the random neural network using quasi-Newton methods
Training in the random neural network (RNN) is generally speci®ed as the minimization of an appropriate error function with respect to the parameters of the network (weights corresponding to positive and negative connections). We propose here a technique for error minimization that is based on the use of quasi-Newton optimization techniques. Such techniques oer more sophisticated exploitation ...
متن کاملQuasi - Newton Trust - Region Method
The classical trust-region method for unconstrained minimization can be augmented with a line search that finds a point that satisfies the Wolfe conditions. One can use this new method to define an algorithm that simultaneously satisfies the quasi-Newton condition at each iteration and maintains a positive-definite approximation to the Hessian of the objective function. This new algorithm has s...
متن کاملAdaptive Neural Network Method for Consensus Tracking of High-Order Mimo Nonlinear Multi-Agent Systems
This paper is concerned with the consensus tracking problem of high order MIMO nonlinear multi-agent systems. The agents must follow a leader node in presence of unknown dynamics and uncertain external disturbances. The communication network topology of agents is assumed to be a fixed undirected graph. A distributed adaptive control method is proposed to solve the consensus problem utilizing re...
متن کاملA parameterized Newton method and a quasi-Newton method for nonsmooth equations
This paper presents a parameterized Newton method using generalized Jacobians and a Broyden-like method for solving nonsmooth equations. The former ensures that the method is well-deened even when the generalized Jacobian is singular. The latter is constructed by using an approximation function which can be formed for nonsmooth equations arising from partial diierential equations and nonlinear ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Nonlinear Theory and Its Applications, IEICE
سال: 2022
ISSN: ['2185-4106']
DOI: https://doi.org/10.1587/nolta.13.361